Goto

Collaborating Authors

 new concept




ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and Acquisition at Inference Time

Neural Information Processing Systems

Humans have the remarkable ability to recognize and acquire novel visual concepts in a zero-shot manner. Given a high-level, symbolic description of a novel concept in terms of previously learned visual concepts and their relations, humans can recognize novel concepts without seeing any examples. Moreover, they can acquire new concepts by parsing and communicating symbolic structures using learned visual concepts and relations. Endowing these capabilities in machines is pivotal in improving their generalization capability at inference time. In this work, we introduce Zero-shot Concept Recognition and Acquisition (ZeroC), a neuro-symbolic architecture that can recognize and acquire novel concepts in a zero-shot way.


Saying the Unsaid: Revealing the Hidden Language of Multimodal Systems Through Telephone Games

Zhao, Juntu, Zhang, Jialing, Li, Chongxuan, Wang, Dequan

arXiv.org Artificial Intelligence

Recent closed-source multimodal systems have made great advances, but their hidden language for understanding the world remains opaque because of their black-box architectures. In this paper, we use the systems' preference bias to study their hidden language: During the process of compressing the input images (typically containing multiple concepts) into texts and then reconstructing them into images, the systems' inherent preference bias introduces specific shifts in the outputs, disrupting the original input concept co-occurrence. We employ the multi-round "telephone game" to strategically leverage this bias. By observing the co-occurrence frequencies of concepts in telephone games, we quantitatively investigate the concept connection strength in the understanding of multimodal systems, i.e., "hidden language." We also contribute Telescope, a dataset of 10,000+ concept pairs, as the database of our telephone game framework. Our telephone game is test-time scalable: By iteratively running telephone games, we can construct a global map of concept connections in multimodal systems' understanding. Here we can identify preference bias inherited from training, assess generalization capability advancement, and discover more stable pathways for fragile concept connections. Furthermore, we use Reasoning-LLMs to uncover unexpected concept relationships that transcend textual and visual similarities, inferring how multimodal systems understand and simulate the world. This study offers a new perspective on the hidden language of multimodal systems and lays the foundation for future research on the interpretability and controllability of multimodal systems.


Explainable Benchmarking through the Lense of Concept Learning

Zhang, Quannian, Röder, Michael, Srivastava, Nikit, Kouagou, N'Dah Jean, Ngomo, Axel-Cyrille Ngonga

arXiv.org Artificial Intelligence

Evaluating competing systems in a comparable way, i.e., benchmarking them, is an undeniable pillar of the scientific method. However, system performance is often summarized via a small number of metrics. The analysis of the evaluation details and the derivation of insights for further development or use remains a tedious manual task with often biased results. Thus, this paper argues for a new type of benchmarking, which is dubbed explainable benchmarking. The aim of explainable benchmarking approaches is to automatically generate explanations for the performance of systems in a benchmark. We provide a first instantiation of this paradigm for knowledge-graph-based question answering systems. We compute explanations by using a novel concept learning approach developed for large knowledge graphs called PruneCEL. Our evaluation shows that PruneCEL outperforms state-of-the-art concept learners on the task of explainable benchmarking by up to 0.55 points F1 measure. A task-driven user study with 41 participants shows that in 80\% of the cases, the majority of participants can accurately predict the behavior of a system based on our explanations. Our code and data are available at https://github.com/dice-group/PruneCEL/tree/K-cap2025




Babies' brains 'tick' more slowly than ours, which may help them learn

New Scientist

Babies' brains'tick' more slowly than ours, which may help them learn The rhythm of an infant's brain activity seems to put them in constant learning mode, whereas that of an adult may allow them to retrieve conceptual knowledge Babies' brains operate at a different rhythm to those of adults When a baby tries to make sense of what they have seen, their brain activity seems to tick at a slower rhythm than it does in adults, which may help them to continually learn new concepts. Our brain processes sensory stimuli using networks of neurons. If a neuron receives a strong enough signal from another neuron, it transmits the signal to more neurons still, producing synchronised waves of electrical activity where many neurons alternate between activated and silent states. Such brainwaves occur at various frequencies. When a given brain region displays a range of frequencies simultaneously, a higher proportion of its neurons may synchronise with certain frequencies more than others.


Identifying Emerging Concepts in Large Corpora

Ma, Sibo, Nyarko, Julian

arXiv.org Artificial Intelligence

We introduce a new method to identify emerging concepts in large text corpora. By analyzing changes in the heatmaps of the underlying embedding space, we are able to detect these concepts with high accuracy shortly after they originate, in turn outperforming common alternatives. We further demonstrate the utility of our approach by analyzing speeches in the U.S. Senate from 1941 to 2015. Our results suggest that the minority party is more active in introducing new concepts into the Senate discourse. We also identify specific concepts that closely correlate with the Senators' racial, ethnic, and gender identities. An implementation of our method is publicly available.


Few-Shot Task Learning through Inverse Generative Modeling

Netanyahu, Aviv, Du, Yilun, Bronars, Antonia, Pari, Jyothish, Tenenbaum, Joshua, Shu, Tianmin, Agrawal, Pulkit

arXiv.org Artificial Intelligence

Learning the intents of an agent, defined by its goals or motion style, is often extremely challenging from just a few examples. We refer to this problem as task concept learning and present our approach, Few-Shot Task Learning through Inverse Generative Modeling (FTL-IGM), which learns new task concepts by leveraging invertible neural generative models. The core idea is to pretrain a generative model on a set of basic concepts and their demonstrations. Then, given a few demonstrations of a new concept (such as a new goal or a new action), our method learns the underlying concepts through backpropagation without updating the model weights, thanks to the invertibility of the generative model. We evaluate our method in five domains -- object rearrangement, goal-oriented navigation, motion caption of human actions, autonomous driving, and real-world table-top manipulation. Our experimental results demonstrate that via the pretrained generative model, we successfully learn novel concepts and generate agent plans or motion corresponding to these concepts in (1) unseen environments and (2) in composition with training concepts.